Neural networks require a careful design in order to perform properly on a given task. In particular, selecting a good activation function (possibly in a data-dependent fashion) is a crucial step, which remains an open problem in the research community. Despite a large amount of investigations, most current implementations simply select one fixed function from a small set of candidates, which is not adapted during training, and is shared among all neurons throughout the different layers. However, neither two of these assumptions can be supposed optimal in practice. In this paper, we present a principled way to have data-dependent adaptation of the activation functions, which is performed independently for each neuron. This is achieved by leveraging over past and present advances on cubic spline interpolation, allowing for local adaptation of the functions around their regions of use. The resulting algorithm is relatively cheap to implement, and overfitting is counterbalanced by the inclusion of a novel damping criterion, which penalizes unwanted oscillations from a predefined shape. Preliminary experimental results validate the proposal.

Learning activation functions from data using cubic spline interpolation / Scardapane, Simone; Scarpiniti, Michele; Comminiello, Danilo; Uncini, Aurelio. - (2019), pp. 73-83. - SMART INNOVATION, SYSTEMS AND TECHNOLOGIES. [10.1007/978-3-319-95098-3_7].

Learning activation functions from data using cubic spline interpolation

Scardapane, Simone
;
Scarpiniti, Michele;Comminiello, Danilo;Uncini, Aurelio
2019

Abstract

Neural networks require a careful design in order to perform properly on a given task. In particular, selecting a good activation function (possibly in a data-dependent fashion) is a crucial step, which remains an open problem in the research community. Despite a large amount of investigations, most current implementations simply select one fixed function from a small set of candidates, which is not adapted during training, and is shared among all neurons throughout the different layers. However, neither two of these assumptions can be supposed optimal in practice. In this paper, we present a principled way to have data-dependent adaptation of the activation functions, which is performed independently for each neuron. This is achieved by leveraging over past and present advances on cubic spline interpolation, allowing for local adaptation of the functions around their regions of use. The resulting algorithm is relatively cheap to implement, and overfitting is counterbalanced by the inclusion of a novel damping criterion, which penalizes unwanted oscillations from a predefined shape. Preliminary experimental results validate the proposal.
2019
Smart Innovation, Systems and Technologies
978-3-319-95097-6
978-3-319-95098-3
Activation function; neural network; spline interpolation; Decision Sciences (all); Computer Science (all)
02 Pubblicazione su volume::02a Capitolo o Articolo
Learning activation functions from data using cubic spline interpolation / Scardapane, Simone; Scarpiniti, Michele; Comminiello, Danilo; Uncini, Aurelio. - (2019), pp. 73-83. - SMART INNOVATION, SYSTEMS AND TECHNOLOGIES. [10.1007/978-3-319-95098-3_7].
File allegati a questo prodotto
File Dimensione Formato  
Scardapane_Learning-activation-functions_2019.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 223.66 kB
Formato Adobe PDF
223.66 kB Adobe PDF   Contatta l'autore

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1241930
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 14
  • ???jsp.display-item.citation.isi??? ND
social impact